Overtraining in neural networks that interpret clinical data.
نویسندگان
چکیده
Backpropagation neural networks are a computer-based pattern-recognition method that has been applied to the interpretation of clinical data. Unlike rule-based pattern recognition, backpropagation networks learn by being repetitively trained with examples of the patterns to be differentiated. We describe and analyze the phenomenon of overtraining in backpropagation networks. Overtraining refers to the reduction in generalization ability that can occur as networks are trained. The clinical application we used was the differentiation of giant cell arteritis (GCA) from other forms of vasculitis (OTH) based on results for 807 patients (593 OTH, 214 GCA) and eight clinical predictor variables. The 807 cases were randomly assigned to either a training set with 404 cases or to a cross-validation set with the remaining 403 cases. The cross-validation set was used to monitor generalization during training. Results were obtained for eight networks, each derived from a different random assignment of the 807 cases. Training error monotonically decreased during training. In contrast, the cross-validation error usually reached a minimum early in training while the training error was still decreasing. Training beyond the minimum cross-validation error was associated with an increased cross-validation error. The shape of the cross-validation error curve and the point during training corresponding to the minimum cross-validation error varied with the composition of the data sets and the training conditions. The study indicates that training error is not a reliable indicator of a network's ability to generalize. To find the point during training when a network generalizes best, one must monitor cross-validation error separately.
منابع مشابه
Dynamics of Batch Learning in Multilayer Networks { Overrealizability and Overtraining {
This paper investigates the dynamics of batch learning of multilayer neural networks in the asymptotic case where the number of training data is much larger than the number of parameters. We consider regression problems assuming noisy output data. First, we present experimental results on the behavior in the steepest descent learning of multilayer per-ceptrons and three-layer linear neural netw...
متن کاملArtificial neural networks: applications in predicting pancreatitis survival
Artificial neural networks are intelligent systems that have successfully been used for prediction in different medical fields. In this study, the efficiency of a neural network for predicting the survival of patients with acute pancreatitis is compared with days-of-survival obtained from patients. A three- layer back-propagation neural network was developed for this purpose. Clinical data (e.g...
متن کاملArtificial neural networks: applications in predicting pancreatitis survival
Artificial neural networks are intelligent systems that have successfully been used for prediction in different medical fields. In this study, the efficiency of a neural network for predicting the survival of patients with acute pancreatitis is compared with days-of-survival obtained from patients. A three- layer back-propagation neural network was developed for this purpose. Clinical data (e.g...
متن کاملLearning in Neural Networks and an Integrable System
This paper investigates the dynamics of batch learning of multilayer neural networks in the asymptotic case where the number of training data is much larger than the number of parameters. First, we present experimental results on the behavior in the steepest descent learning of multilayer perceptrons and three-layer linear neural networks. We see in these results that strong overtraining, which...
متن کاملDynamics of Batch Learning in Multilayer Neural Networks
We discuss the dynamics of batch learning of multilayer neural networks in the asymptotic limit, where the number of trining data is much larger than the number of parameters, emphasizing on the parameterization redundancy in overrealizable cases. In addition to showing experimental results on overtraining in multilayer perceptrons and three-layer linear neural networks, we theoretically prove ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Clinical chemistry
دوره 39 9 شماره
صفحات -
تاریخ انتشار 1993